Analysis of stochastic gradient descent in continuous time

نویسندگان

چکیده

Abstract Stochastic gradient descent is an optimisation method that combines classical with random subsampling within the target functional. In this work, we introduce stochastic process as a continuous-time representation of descent. The dynamical system coupled Markov living on finite state space. system—a flow—represents part, space represents subsampling. Processes type are, for instance, used to model clonal populations in fluctuating environments. After introducing it, study theoretical properties process: We show it converges weakly flow respect full function, learning rate approaches zero. give conditions under which constant exponentially ergodic Wasserstein sense. Then case, where goes zero sufficiently slowly and single functions are strongly convex. point mass concentrated global minimum function; indicating consistency method. conclude after discussion discretisation strategies numerical experiments.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Stochastic Gradient Descent in Continuous Time

We consider stochastic gradient descent for continuous-time models. Traditional approaches for the statistical estimation of continuous-time models, such as batch optimization, can be impractical for large datasets where observations occur over a long period of time. Stochastic gradient descent provides a computationally efficient method for such statistical learning problems. The stochastic gr...

متن کامل

Continuous-Time Limit of Stochastic Gradient Descent Revisited

Stochastic Gradient Descent (SGD) is an important algorithm in machine learning. With constant learning rates, it is a stochastic process that reaches a stationary distribution. We revisit an analysis of SGD in terms of stochastic differential equations in the limit of small constant gradient steps. This limit, which we feel is not appreciated in the machine learning community, allows us to app...

متن کامل

Convergence Analysis of Gradient Descent Stochastic Algorithms

This paper proves convergence of a sample-path based stochastic gradient-descent algorithm for optimizing expected-value performance measures in discrete event systems. The algorithm uses increasing precision at successive iterations, and it moves against the direction of a generalized gradient of the computed sample performance function. Two convergence results are established: one, for the ca...

متن کامل

Continuous Generalized Gradient Descent

This article derives characterizations and computational algorithms for continuous general gradient descent trajectories in high-dimensional parameter spaces for statistical model selection, prediction, and classification. Examples include proportional gradient shrinkage as an extension of LASSO and LARS, threshold gradient descent with right-continuous variable selectors, threshold ridge regre...

متن کامل

Hilbert maps: scalable continuous occupancy mapping with stochastic gradient descent

The vast amount of data robots can capture today motivates the development of fast and scalable statistical tools to model the environment the robot operates in. We devise a new technique for environment representation through continuous occupancy mapping that improves on the popular occupancy grip maps in two fundamental aspects: 1) it does not assume an a priori discretisation of the world in...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Statistics and Computing

سال: 2021

ISSN: ['0960-3174', '1573-1375']

DOI: https://doi.org/10.1007/s11222-021-10016-8